Parallel stochastic gradient algorithms for large-scale matrix completion

نویسندگان

  • Benjamin Recht
  • Christopher Ré
چکیده

This paper develops Jellyfish, an algorithm for solving data-processing problems with matrix-valued decision variables regularized to have low rank. Particular examples of problems solvable by Jellyfish include matrix completion problems and least-squares problems regularized by the nuclear norm or γ2-norm. Jellyfish implements a projected incremental gradient method with a biased, random ordering of the increments. This biased ordering allows for a parallel implementation that admits a speed-up nearly proportional to the number of processors. On large-scale matrix completion tasks, Jellyfish is orders of magnitude more efficient than existing codes. For example, on the Netflix Prize data set, prior art computes rating predictions in approximately 4 hours, while Jellyfish solves the same problem in under 3 minutes on a 12 core workstation.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

GPUFish: A Parallel Computing Framework for Matrix Completion from A Few Observations

The problem of recovering a data matrix from a small sample of observed entries, also known as matrix completion, arises in several real-world applications including recommender systems, sensor localization, and system identification. We introduce GPUFish, a parallel computing software framework for solving very large-scale matrix completion problems. GPUFish is modular, tunable, inherently par...

متن کامل

Online Passive-Aggressive Algorithms for Non-Negative Matrix Factorization and Completion

Stochastic Gradient Descent (SGD) is a popular online algorithm for large-scale matrix factorization. However, SGD can often be di cult to use for practitioners, because its performance is very sensitive to the choice of the learning rate parameter. In this paper, we present non-negative passiveaggressive (NN-PA), a family of online algorithms for non-negative matrix factorization (NMF). Our al...

متن کامل

Tensor Completion

The purpose of this thesis is to explore the methods to solve the tensor completion problem. Inspired by the matrix completion problem, the tensor completion problem is formulated as an unconstrained nonlinear optimization problem, which finds three factors that give a low-rank approximation. Various of iterative methods, including the gradient-based methods, stochastic gradient descent method ...

متن کامل

A Saddle Point Approach to Structured Low-rank Matrix Learning in Large-scale Applications

We propose a novel optimization approach for learning a low-rank matrix which is also constrained to lie in a linear subspace. Exploiting a particular variational characterization of the squared trace norm regularizer, we formulate the structured low-rank matrix learning problem as a rank-constrained saddle point minimax problem. The proposed modeling decouples the lowrank and structural constr...

متن کامل

Optimal Discrete Matrix Completion

In recent years, matrix completion methods have been successfully applied to solve recommender system applications. Most of them focus on the matrix completion problem in real number domain, and produce continuous prediction values. However, these methods are not appropriate in some occasions where the entries of matrix are discrete values, such as movie ratings prediction, social network relat...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:
  • Math. Program. Comput.

دوره 5  شماره 

صفحات  -

تاریخ انتشار 2013